Goto

Collaborating Authors

 outlier-robust estimation


Outlier-robust estimation of a sparse linear model using \ell_1 -penalized Huber's M -estimator

Neural Information Processing Systems

We study the problem of estimating a $p$-dimensional $s$-sparse vector in a linear model with Gaussian design. In the case where the labels are contaminated by at most $o$ adversarial outliers, we prove that the $\ell_1$-penalized Huber's $M$-estimator based on $n$ samples attains the optimal rate of convergence $(s/n)^{1/2} + (o/n)$, up to a logarithmic factor. For more general design matrices, our results highlight the importance of two properties: the transfer principle and the incoherence property. These properties with suitable constants are shown to yield the optimal rates of robust estimation with adversarial contamination.


Reviews: Outlier-robust estimation of a sparse linear model using \ell_1 -penalized Huber's M -estimator

Neural Information Processing Systems

This is a very good paper. I really enjoyed reading the paper. The result is very strong and the paper is very readable. I strongly recommend accepting the paper, if the proof is correct. This paper solves an important open problem in robust estimation.


Outlier-robust estimation of a sparse linear model using \ell_1 -penalized Huber's M -estimator

Neural Information Processing Systems

We study the problem of estimating a p -dimensional s -sparse vector in a linear model with Gaussian design. In the case where the labels are contaminated by at most o adversarial outliers, we prove that the \ell_1 -penalized Huber's M -estimator based on n samples attains the optimal rate of convergence (s/n) {1/2} (o/n), up to a logarithmic factor. For more general design matrices, our results highlight the importance of two properties: the transfer principle and the incoherence property. These properties with suitable constants are shown to yield the optimal rates of robust estimation with adversarial contamination.


Outlier-robust estimation of a sparse linear model using \ell_1-penalized Huber's M-estimator

Dalalyan, Arnak, Thompson, Philip

Neural Information Processing Systems

We study the problem of estimating a $p$-dimensional $s$-sparse vector in a linear model with Gaussian design. In the case where the labels are contaminated by at most $o$ adversarial outliers, we prove that the $\ell_1$-penalized Huber's $M$-estimator based on $n$ samples attains the optimal rate of convergence $(s/n) {1/2} (o/n)$, up to a logarithmic factor. For more general design matrices, our results highlight the importance of two properties: the transfer principle and the incoherence property. These properties with suitable constants are shown to yield the optimal rates of robust estimation with adversarial contamination. Papers published at the Neural Information Processing Systems Conference.